Goto

Collaborating Authors

 hierarchical structure




Variational Temporal Abstraction

Taesup Kim, Sungjin Ahn, Yoshua Bengio

Neural Information Processing Systems

There have been approaches to learn such hierarchical structure in sequences such as the HMRNN (Chung et al., 2016). However, as a deterministic model, it has the main limitation that it cannot capture the stochastic nature prevailing in the data. In particular,this is acritical limitation to imagination-augmented agents because exploring various possible futures according to the uncertainty is what makes the imagination meaningful in many cases.




Novel positional encodings to enable tree-based transformers

Vighnesh Shiv, Chris Quirk

Neural Information Processing Systems

Motivated by this property, we propose a method to extend transformers to tree-structured data, enabling sequence-totree, tree-to-sequence, and tree-to-tree mappings. Our approach abstracts the transformer'ssinusoidal positional encodings, allowing ustoinstead useanovel positional encoding scheme to represent node positions within trees.